301 research outputs found

    Differentially Private Publication of Sparse Data

    Full text link
    The problem of privately releasing data is to provide a version of a dataset without revealing sensitive information about the individuals who contribute to the data. The model of differential privacy allows such private release while providing strong guarantees on the output. A basic mechanism achieves differential privacy by adding noise to the frequency counts in the contingency tables (or, a subset of the count data cube) derived from the dataset. However, when the dataset is sparse in its underlying space, as is the case for most multi-attribute relations, then the effect of adding noise is to vastly increase the size of the published data: it implicitly creates a huge number of dummy data points to mask the true data, making it almost impossible to work with. We present techniques to overcome this roadblock and allow efficient private release of sparse data, while maintaining the guarantees of differential privacy. Our approach is to release a compact summary of the noisy data. Generating the noisy data and then summarizing it would still be very costly, so we show how to shortcut this step, and instead directly generate the summary from the input data, without materializing the vast intermediate noisy data. We instantiate this outline for a variety of sampling and filtering methods, and show how to use the resulting summary for approximate, private, query answering. Our experimental study shows that this is an effective, practical solution, with comparable and occasionally improved utility over the costly materialization approach

    One-loop expressions for hllˉγh\rightarrow l\bar{l}\gamma in Higgs extensions of the Standard Model

    Full text link
    A systematic study of one-loop contributions to the decay channels hllˉγh\rightarrow l\bar{l}\gamma with l=νe,μ,τ,e,μl=\nu_{e,\mu, \tau}, e, \mu, performed in Higgs extended versions of the Standard Model, is presented in the 't Hooft-Veltman gauge. Analytic formulas for one-loop form factors are expressed in terms of the logarithm and di-logarithmic functions. As a result, these form factors can be reduced to those relating to the loop-induced decay processes hγγ,Zγh\rightarrow \gamma\gamma, Z\gamma, confirming not only previous results using different approaches but also close relations between the three kinds of the loop-induced Higgs decay rates. For phenomenological study, we focus on the two observables, namely the enhancement factors defined as ratios of the decay rates calculated between the Higgs extended versions and the standard model, and the forward-backward asymmetries of fermions, which can be used to search for Higgs extensions of the SM. We show that direct effects of mixing between neutral Higgs bosons and indirect contributions of charged Higg boson exchanges can be probed at future colliders.Comment: 39 pages, 9 Figures, 11 Tables of dat

    Assessment of seasonal winter temperature forecast errors in the regcm model over northern Vietnam

    Get PDF
    This study verified the seasonal six-month forecasts for winter temperatures for northern Vietnam in 1998–2018 using a regional climate model (RegCM4) with the boundary conditions of the climate forecast system Version 2 (CFSv2) from the National Centers for Environmental Prediction (NCEP). First, different physical schemes (land-surface process, cumulus, and radiation parameterizations) in RegCM4 were applied to generate 12 single forecasts. Second, the simple ensemble forecasts were generated through the combinations of those different physical formulations. Three subclimate regions (R1, R2, R3) of northern Vietnam were separately tested with surface observations and a reanalysis dataset (Japanese 55-year reanalysis (JRA55)). The highest sensitivity to the mean monthly temperature forecasts was shown by the land-surface parameterizations (the biosphere−atmosphere transfer scheme (BATS) and community land model version 4.5 (CLM)). The BATS forecast groups tended to provide forecasts with lower temperatures than the actual observations, while the CLM forecast groups tended to overestimate the temperatures. The forecast errors from single forecasts could be clearly reduced with ensemble mean forecasts, but ensemble spreads were less than those root-mean-square errors (RMSEs). This indicated that the ensemble forecast was underdispersed and that the direct forecast from RegCM4 needed more postprocessing

    Seasonal variation of phytoplankton in My Thanh River, Mekong delta, Vietnam

    Get PDF
    A study on the seasonal variation of phytoplankton composition was conducted at the upper, middle, and lower parts of the My Thanh River, which supplies an important source of water for aquaculture. Qualitative and quantitative samples of phytoplankton were collected monthly at both high and low tide. The results showed that a total of 171 phytoplankton (algae) species were recorded, belonging to 59 genera and 5 phyla. Diatoms were the most abundant group with the highest species number, followed by green algae. The other phyla possessed a lower number of species. The species composition was more diverse in the rainy season and at high tide at most of the sampling sites. The mean density of algae varied from 30,900-43,521 ind.L^-1^. The density of diatoms was higher in the middle and lower parts. At the same time, euglenoids displayed the highest density in the upper part, showing a difference in the dominant algae group under the influence of salinity. Salinity was found to be significantly positively correlated (p<0.01) with diatoms, whereas it was negatively correlated (p<0.05) with blue-green algae and euglenoids. The algae composition was quite diverse, with the H' index ranging from 2.0-3.3, showing the water quality was slightly to moderately polluted

    Discovering Valuable Items from Massive Data

    Full text link
    Suppose there is a large collection of items, each with an associated cost and an inherent utility that is revealed only once we commit to selecting it. Given a budget on the cumulative cost of the selected items, how can we pick a subset of maximal value? This task generalizes several important problems such as multi-arm bandits, active search and the knapsack problem. We present an algorithm, GP-Select, which utilizes prior knowledge about similarity be- tween items, expressed as a kernel function. GP-Select uses Gaussian process prediction to balance exploration (estimating the unknown value of items) and exploitation (selecting items of high value). We extend GP-Select to be able to discover sets that simultaneously have high utility and are diverse. Our preference for diversity can be specified as an arbitrary monotone submodular function that quantifies the diminishing returns obtained when selecting similar items. Furthermore, we exploit the structure of the model updates to achieve an order of magnitude (up to 40X) speedup in our experiments without resorting to approximations. We provide strong guarantees on the performance of GP-Select and apply it to three real-world case studies of industrial relevance: (1) Refreshing a repository of prices in a Global Distribution System for the travel industry, (2) Identifying diverse, binding-affine peptides in a vaccine de- sign task and (3) Maximizing clicks in a web-scale recommender system by recommending items to users

    Resource Competition on Integral Polymatroids

    Full text link
    We study competitive resource allocation problems in which players distribute their demands integrally on a set of resources subject to player-specific submodular capacity constraints. Each player has to pay for each unit of demand a cost that is a nondecreasing and convex function of the total allocation of that resource. This general model of resource allocation generalizes both singleton congestion games with integer-splittable demands and matroid congestion games with player-specific costs. As our main result, we show that in such general resource allocation problems a pure Nash equilibrium is guaranteed to exist by giving a pseudo-polynomial algorithm computing a pure Nash equilibrium.Comment: 17 page

    Conquering hypertension in Vietnam-solutions at grassroots level: study protocol of a cluster randomized controlled trial

    Get PDF
    BACKGROUND: Vietnam has been experiencing an epidemiologic transition to that of a lower-middle income country with an increasing prevalence of non-communicable diseases. The key risk factors for cardiovascular disease (CVD) are either on the rise or at alarming levels in Vietnam, particularly hypertension (HTN). Inasmuch, the burden of CVD will continue to increase in the Vietnamese population unless effective prevention and control measures are put in place. The objectives of the proposed project are to evaluate the implementation and effectiveness of two multi-faceted community and clinic-based strategies on the control of elevated blood pressure (BP) among adults in Vietnam via a cluster randomized trial design. METHODS: Sixteen communities will be randomized to either an intervention (8 communities) or a comparison group (8 communities). Eligible and consenting adult study participants with HTN (n = 680) will be assigned to intervention/comparison status based on the community in which they reside. Both comparison and intervention groups will receive a multi-level intervention modeled after the Vietnam National Hypertension Program including education and practice change modules for health care providers, accessible reading materials for patients, and a multi-media community awareness program. In addition, the intervention group only will receive three carefully selected enhancements integrated into routine clinical care: (1) expanded community health worker services, (2) home BP self-monitoring, and (3) a storytelling intervention, which consists of interactive, literacy-appropriate, and culturally sensitive multi-media storytelling modules for motivating behavior change through the power of patients speaking in their own voices. The storytelling intervention will be delivered by DVDs with serial installments at baseline and at 3, 6, and 9 months after trial enrollment. Changes in BP will be assessed in both groups at several follow-up time points. Implementation outcomes will be assessed as well. DISCUSSION: Results from this full-scale trial will provide health policymakers with practical evidence on how to combat a key risk factor for CVD using a feasible, sustainable, and cost-effective intervention that could be used as a national program for controlling HTN in Vietnam. TRIAL REGISTRATION: ClinicalTrials.gov NCT03590691 . Registered on July 17, 2018. Protocol version: 6. Date: August 15, 2019

    Solar models and solar neutrino oscillations

    Get PDF
    We provide a summary of the current knowledge, theoretical and experimental, of solar neutrino fluxes and of the masses and mixing angles that characterize solar neutrino oscillations. We also summarize the principal reasons for doing new solar neutrino experiments and what we think may be learned from the future measurements.Comment: Submitted to the Neutrino Focus Issue of New Journal of Physics at http://www.njp.or

    Mapping for engagement: setting up a community based participatory research project to reach underserved communities at risk for Hepatitis C in Ho Chi Minh City, Vietnam

    Get PDF
    Background: Approximately 1. 07 million people in Vietnam are infected with hepatitis C virus (HCV). To address this epidemic, the South East Asian Research Collaborative in Hepatitis (SEARCH) launched a 600-patient cohort study and two clinical trials, both investigating shortened treatment strategies for chronic HCV infection with direct-acting antiviral drugs. We conducted ethnographic research with a subset of trial participants and found that the majority were aware of HCV infection and its implications and were motivated to seek treatment. However, people who inject drugs (PWID), and other groups at risk for HCV were under-represented, although injecting drug use is associated with high rates of HCV. Material and Methods: We designed a community-based participatory research (CBPR) study to engage in dialogues surrounding HCV and other community-prioritized health issues with underserved groups at risk for HCV in Ho Chi Minh City. The project consists of three phases: situation analysis, CBPR implementation, and dissemination. In this paper, we describe the results of the first phase (i.e., the situation analysis) in which we conducted desk research and organized stakeholder mapping meetings with representatives from local non-government and community-based organizations where we used participatory research methods to identify and analyze key stakeholders working with underserved populations. Results: Twenty six institutions or groups working with the key underserved populations were identified. Insights about the challenges and dynamics of underserved communities were also gathered. Two working groups made up of representatives from the NGO and CBO level were formed. Discussion: Using the information provided by local key stakeholders to shape the project has helped us to build solid relationships, give the groups a sense of ownership from the early stages, and made the project more context specific. These steps are not only important preliminary steps for participatory studies but also for other research that takes place within the communities
    corecore